Machine Translation ((MT)) refers to the automated software-based Translation of natural language text. The embedded complexities and incompatibilities of natural languages have made (MT) a daunting task facing numerous challenges, especially when it is to be compared to a manual Translation. With the emergence of deep-learning AI approaches, the Neural Machine Translation (N(MT)) has pushed (MT) results closer to human expectations. One of the newest deep learning approaches is the sequence-to-sequence approach based on Recurrent Neural Networks (RNN), complex convolutions, and transformers, and employing encoders/decoder pairs. In this study, an attention-based deep learning architecture is proposed for (MT), with all layers focused exclusively on multi-head attention and based on a transformer that includes multi-layer encoders/decoders. The main contributions of the proposed model lie in the weighted combination of layers’ primary input and output of the previous layers, feeding into the next layer. This mechanism results in a more accurate transformation compared to nonhybrid inputs. The model is evaluated using two datasets for German/English Translation, the W(MT)'14 dataset for training, and the newstest'2012 dataset for testing. The experiments are run on GPU-equipped Google Colab instances and the results show an accuracy of 36. 7 BLEU, a 5% improvement over the previous work without the hybrid-input technique.